Adaptive Prior-Dependent Correction Enhanced Reinforcement Learning for Natural Language Generation
نویسندگان
چکیده
Natural language generation (NLG) is an important task with various applications like neural machine translation (NMT) and image captioning. Since deep-learning-based methods have issues of exposure bias loss inconsistency, reinforcement learning (RL) widely adopted in NLG tasks recently. But most RL-based ignore the deviation ignorance issue, which means model fails to understand extent token-level well. It leads semantic incorrectness hampers agent perform To address we propose a technique called adaptive prior-dependent correction (APDC) enhance RL. leverages distribution generated by computing distances between ground truth all other words correct agent's stochastic policy. Additionally, some techniques on RL are explored coordinate APDC, requires reward estimation at every time step. We find that special case RL, where state transition deterministic afterstate value equals Q-value utilize such prior knowledge, estimate advantage function difference Q-values can be estimated Monte Carlo rollouts. Experiments show that, three (NMT, captioning, abstractive text summarization), our method consistently outperforms state-of-the-art approaches different frequently-used metrics.
منابع مشابه
Adaptive Natural Language Generation in Dialogue using Reinforcement Learning
This paper presents a new model for adaptive Natural Language Generation (NLG) in dialogue, showing how NLG problems can be approached as statistical planning problems using Reinforcement Learning. This approach brings a number of theoretical and practical benefits such as finegrained adaptation, generalization, and automatic (global) optimization. We present the model and related work in stati...
متن کاملHierarchical reinforcement learning for situated natural language generation
Natural Language Generation systems in interactive settings often face a multitude of choices, given that the communicative effect of each utterance they generate depends crucially on the interplay between its physical circumstances, addressee and interaction history. This is particularly true in interactive and situated settings. In this paper we present a novel approach for situated Natural L...
متن کاملDeep Reinforcement Learning for Programming Language Correction
Novice programmers often struggle with the formal syntax of programming languages. To assist them, we design a novel programming language correction framework amenable to reinforcement learning. The framework allows an agent to mimic human actions for text navigation and editing. We demonstrate that the agent can be trained through self-exploration directly from the raw input, that is, program ...
متن کاملNatural Language Generation as Planning under Uncertainty Using Reinforcement Learning
We present and evaluate a new model for Natural Language Generation (NLG) in Spoken Dialogue Systems, based on statistical planning, given noisy feedback from the current generation context (e.g. a user and a surface realiser). We study its use in a standard NLG problem: how to present information (in this case a set of search results) to users, given the complex tradeoffs between utterance len...
متن کاملHierarchical Reinforcement Learning for Adaptive Text Generation
We present a novel approach to natural language generation (NLG) that applies hierarchical reinforcement learning to text generation in the wayfinding domain. Our approach aims to optimise the integration of NLG tasks that are inherently different in nature, such as decisions of content selection, text structure, user modelling, referring expression generation (REG), and surface realisation. It...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2021
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v35i14.17504